首页> 外文OA文献 >On continuous space word representations as input of LSTM language model
【2h】

On continuous space word representations as input of LSTM language model

机译:以连续的空间词表示形式作为LSTM语言模型的输入

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Artificial neural networks have become the state-of-the-art in the task of language modelling whereas Long-Short Term Memory (LSTM) networks seem to be an efficient architecture. The continuous skip-gram and the continuous bag of words (CBOW) are algorithms for learning quality distributed vector representations that are able to capture a large number of syntactic and semantic word relationships. In this paper, we carried out experiments with a combination of these powerful models: the continuous representations of words trained with skip-gram/CBOW/GloVe method, word cache expressed as a vector using latent Dirichlet allocation (LDA). These all are used on the input of LSTM network instead of 1-of-N coding traditionally used in language models. The proposed models are tested on Penn Treebank and MALACH corpus.
机译:人工神经网络已成为语言建模任务中的最新技术,而长短期记忆(LSTM)网络似乎是一种有效的体系结构。连续跳跃语法和连续词袋(CBOW)是用于学习能够捕获大量句法和语义词关系的质量分布矢量表示的算法。在本文中,我们结合了这些强大的模型进行了实验:使用skip-gram / CBOW / GloVe方法训练的单词的连续表示,使用潜在Dirichlet分配(LDA)表示为矢量的单词缓存。这些全部用于LSTM网络的输入,而不是传统上在语言模型中使用的1-of-N编码。建议的模型在Penn Treebank和MALACH语料库上进行了测试。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号